data modeling
Supplementary Material A Data Modeling
In this section, we provide further details for our data modeling. We note the difficulties of appropriately modeling the terminal variable which is a binary variable compared to the rest of the dimensions which are continuous for the environments we investigate. This is particularly challenging for "expert" datasets where early termination is rare. An immediate advantage of sampling data from a generative model is compression. As we discuss in Appendix B.3, sampling is fast ER provides high levels of dataset compression without sacrificing downstream performance in offline reinforcement learning.
ONION: A Multi-Layered Framework for Participatory ER Design
Makovska, Viktoriia, Fletcher, George, Stoyanovich, Julia
We present ONION, a multi-layered framework for participatory Entity-Relationship (ER) modeling that integrates insights from design justice, participatory AI, and conceptual modeling. ONION introduces a five-stage methodology: Observe, Nurture, Integrate, Optimize, Normalize. It supports progressive abstraction from unstructured stakeholder input to structured ER diagrams. Our approach aims to reduce designer bias, promote inclusive participation, and increase transparency through the modeling process. We evaluate ONION through real-world workshops focused on sociotechnical systems in Ukraine, highlighting how diverse stakeholder engagement leads to richer data models and deeper mutual understanding. Early results demonstrate ONION's potential to host diversity in early-stage data modeling. We conclude with lessons learned, limitations and challenges involved in scaling and refining the framework for broader adoption.
- Europe > Germany > Berlin (0.06)
- North America > United States > New York > New York County > New York City (0.04)
- North America > United States > Hawaii (0.04)
- (3 more...)
CREDAL: Close Reading of Data Models
Fletcher, George, Nahurna, Olha, Prytula, Matvii, Stoyanovich, Julia
Data models are necessary for the birth of data and of any data-driven system. Indeed, every algorithm, every machine learning model, every statistical model, and every database has an underlying data model without which the system would not be usable. Hence, data models are excellent sites for interrogating the (material, social, political, ...) conditions giving rise to a data system. Towards this, drawing inspiration from literary criticism, we propose to closely read data models in the same spirit as we closely read literary artifacts. Close readings of data models reconnect us with, among other things, the materiality, the genealogies, the techne, the closed nature, and the design of technical systems. While recognizing from literary theory that there is no one correct way to read, it is nonetheless critical to have systematic guidance for those unfamiliar with close readings. This is especially true for those trained in the computing and data sciences, who too often are enculturated to set aside the socio-political aspects of data work. A systematic methodology for reading data models currently does not exist. To fill this gap, we present the CREDAL methodology for close readings of data models. We detail our iterative development process and present results of a qualitative evaluation of CREDAL demonstrating its usability, usefulness, and effectiveness in the critical study of data.
- North America > United States > New York (0.04)
- Europe > Ukraine > Lviv Oblast > Lviv (0.04)
- North America > United States > Wisconsin > Dane County > Madison (0.04)
- (5 more...)
- Research Report (1.00)
- Personal > Interview (0.93)
- Information Technology > Information Management (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning (1.00)
- Information Technology > Artificial Intelligence > Cognitive Science > Problem Solving (0.46)
Nd-BiMamba2: A Unified Bidirectional Architecture for Multi-Dimensional Data Processing
Deep learning models often require specially designed architectures to process data of different dimensions, such as 1D time series, 2D images, and 3D volumetric data. Existing bidirectional models mainly focus on sequential data, making it difficult to scale effectively to higher dimensions. To address this issue, we propose a novel multi-dimensional bidirectional neural network architecture, named Nd-BiMamba2, which efficiently handles 1D, 2D, and 3D data. Nd-BiMamba2 is based on the Mamba2 module and introduces innovative bidirectional processing mechanisms and adaptive padding strategies to capture bidirectional information in multi-dimensional data while maintaining computational efficiency. Unlike existing methods that require designing specific architectures for different dimensional data, Nd-BiMamba2 adopts a unified architecture with a modular design, simplifying development and maintenance costs. To verify the portability and flexibility of Nd-BiMamba2, we successfully exported it to ONNX and TorchScript and tested it on different hardware platforms (e.g., CPU, GPU, and mobile devices). Experimental results show that Nd-BiMamba2 runs efficiently on multiple platforms, demonstrating its potential in practical applications. The code is open-source: https://github.com/Human9000/nd-Mamba2-torch
- Europe > Germany > Bavaria > Upper Bavaria > Munich (0.04)
- Asia > China > Shandong Province > Qingdao (0.04)
Unlocking the Diagnostic Potential of ECG through Knowledge Transfer from Cardiac MRI
Turgut, Özgün, Müller, Philip, Hager, Paul, Shit, Suprosanna, Starck, Sophie, Menten, Martin J., Martens, Eimo, Rueckert, Daniel
The electrocardiogram (ECG) is a widely available diagnostic tool that allows for a cost-effective and fast assessment of the cardiovascular health. However, more detailed examination with expensive cardiac magnetic resonance (CMR) imaging is often preferred for the diagnosis of cardiovascular diseases. While providing detailed visualization of the cardiac anatomy, CMR imaging is not widely available due to long scan times and high costs. To address this issue, we propose the first self-supervised contrastive approach that transfers domain-specific information from CMR images to ECG embeddings. Our approach combines multimodal contrastive learning with masked data modeling to enable holistic cardiac screening solely from ECG data. In extensive experiments using data from 40,044 UK Biobank subjects, we demonstrate the utility and generalizability of our method. We predict the subject-specific risk of various cardiovascular diseases and determine distinct cardiac phenotypes solely from ECG data. In a qualitative analysis, we demonstrate that our learned ECG embeddings incorporate information from CMR image regions of interest. We make our entire pipeline publicly available, including the source code and pre-trained model weights.
- Europe > Germany > Bavaria > Upper Bavaria > Munich (0.04)
- Europe > United Kingdom > England > Greater London > London (0.04)
Remote Data Architect openings near you -Updated October 11, 2022 - Remote Tech Jobs
Role requiring'No experience data provided' months of experience in Richmond Pay if you succeed in getting hired and start work at a high-paying job first. Get Paid to Read Emails, Play Games, Search the Web, $5 Signup Bonus. Role requiring'No experience data provided' months of experience in None Pay if you succeed in getting hired and start work at a high-paying job first. Get Paid to Read Emails, Play Games, Search the Web, $5 Signup Bonus. Piper Enterprise Solutions is searching for a Principal Data Architect for a Healthcare Data and Information company.
- North America > United States > Ohio > Hamilton County > Cincinnati (0.04)
- North America > United States > Washington > King County > Bellevue (0.04)
- Law (1.00)
- Information Technology > Security & Privacy (1.00)
- Health & Medicine (1.00)
- Banking & Finance (0.93)
Remote Data Architect openings near you -Updated October 01, 2022 - Remote Tech Jobs
Do you have Data Warehousing, Hadoop/Data Lake experience? Do you like to solve the most complex and high scale (billions records) data challenges in the world today? Do you like to work on-site in a variety of business environments, leading teams through high impact projects that use the newest data analytic technologies? Would you like a career path that enables you to progress with the rapid adoption of cloud computing? This role will specifically focus on large scale data warehousing and data warehouse modernization.
- North America > United States > Colorado (0.04)
- North America > United States > Virginia > Arlington County > Arlington (0.04)
- North America > United States > New York > Erie County > Buffalo (0.04)
- (3 more...)
- Health & Medicine > Therapeutic Area (1.00)
- Government > Regional Government > North America Government > United States Government (1.00)
- Information Technology > Services (0.94)
- Law (0.93)
Graph Modeling in Computer Assisted Automotive Development
Pakiman, Anahita, Garcke, Jochen
We consider graph modeling for a knowledge graph for vehicle development, with a focus on crash safety. An organized schema that incorporates information from various structured and unstructured data sources is provided, which includes relevant concepts within the domain. In particular, we propose semantics for crash computer aided engineering (CAE) data, which enables searchability, filtering, recommendation, and prediction for crash CAE data during the development process. This graph modeling considers the CAE data in the context of the R\&D development process and vehicle safety. Consequently, we connect CAE data to the protocols that are used to assess vehicle safety performances. The R\&D process includes CAD engineering and safety attributes, with a focus on multidisciplinary problem-solving. We describe previous efforts in graph modeling in comparison to our proposal, discuss its strengths and limitations, and identify areas for future work.
- Automobiles & Trucks (1.00)
- Information Technology (0.68)
- Transportation > Ground > Road (0.68)
Principal Data Architect (100% Remote) - Remote Tech Jobs
Piper Enterprise Solutions is searching for a Principal Data Architect for a Healthcare Data and Information company. This position is 100% remote. Qualifications for the Principal Data Architect: • 8 years of engineering or architecture experience with distributed data systems, data mapping, and building ETL/ELT pipelines • Healthcare data (EHR, claims, pharmacy, etc.) is required • 5 years of experience in data modeling and building and modifying in a relational database environment • Minimum of 3 years of experience with data mapping • Expertise in SQL Server • Cloud experience with cloud (AWS or Azure) and Databricks • Programming experience with Scala or Python is a plus
3 Best Practices For Predictive Data Modeling
Predictive modeling is used to develop models that use past occurrences as reference points for organizations to forecast future business-related events and make clever decisions. It is heavily involved in the strategy-making processes of companies in industries such as healthcare, law enforcement, pharmaceuticals and many more. The practices that can be used to make predictive data modeling error-free can be of great importance to everybody. Predictive data modeling involves the creation, testing and validation of data models that will be used for predictive analysis in businesses. The lifecycle management of such models is a part of predictive data modeling.
- Information Technology > Security & Privacy (0.96)
- Health & Medicine (0.89)